Goto

Collaborating Authors

 high resolution image




1. Privacy preserving. R4/R3: the features maps might leak privacy; R1: privacy property has not been described

Neural Information Processing Systems

We will discuss the privacy concerns in our revision. Analyzing the degree of privacy leakage under our framework will be our future works. Please note that the sizes of features and logits are fixed by the CNN architecture design. We will explain communication more obviously in our revision. R2: subsampling can save communication cost." Our method definitely can support the cross-device setting. It is better to tailor the strategy for different models and optimization methods. All clients are divided into many groups. The optimizer state of each group can be maintained by uploading it to the server of GKT. Once the group ID is changed, the server then synchronizes the optimizer state to the clients in the new group. We will demonstrate this in our revision. We can use the edge server in the hierarchical topology to improve load balance. Please provide IID and non-IID experiments for ablation study" The word "diverge" is somewhat misleading.


Simple diffusion: End-to-end diffusion for high resolution images

Hoogeboom, Emiel, Heek, Jonathan, Salimans, Tim

arXiv.org Machine Learning

Currently, applying diffusion models in pixel space of high resolution images is difficult. Instead, existing approaches focus on diffusion in lower dimensional spaces (latent diffusion), or have multiple super-resolution levels of generation referred to as cascades. The downside is that these approaches add additional complexity to the diffusion framework. This paper aims to improve denoising diffusion for high resolution images while keeping the model as simple as possible. The paper is centered around the research question: How can one train a standard denoising diffusion models on high resolution images, and still obtain performance comparable to these alternate approaches? The four main findings are: 1) the noise schedule should be adjusted for high resolution images, 2) It is sufficient to scale only a particular part of the architecture, 3) dropout should be added at specific locations in the architecture, and 4) downsampling is an effective strategy to avoid high resolution feature maps. Combining these simple yet effective techniques, we achieve state-of-the-art on image generation among diffusion models without sampling modifiers on ImageNet.


Using Neural Networks for Fast SAR Roughness Estimation of High Resolution Images

Fan, Li, Neto, Jeova Farias Sales Rocha

arXiv.org Artificial Intelligence

The analysis of Synthetic Aperture Radar (SAR) imagery is an important step in remote sensing applications, and it is a challenging problem due to its inherent speckle noise. One typical solution is to model the data using the $G_I^0$ distribution and extract its roughness information, which in turn can be used in posterior imaging tasks, such as segmentation, classification and interpretation. This leads to the need of quick and reliable estimation of the roughness parameter from SAR data, especially with high resolution images. Unfortunately, traditional parameter estimation procedures are slow and prone to estimation failures. In this work, we proposed a neural network-based estimation framework that first learns how to predict underlying parameters of $G_I^0$ samples and then can be used to estimate the roughness of unseen data. We show that this approach leads to an estimator that is quicker, yields less estimation error and is less prone to failures than the traditional estimation procedures for this problem, even when we use a simple network. More importantly, we show that this same methodology can be generalized to handle image inputs and, even if trained on purely synthetic data for a few seconds, is able to perform real time pixel-wise roughness estimation for high resolution real SAR imagery.


Automatic Cadastral Boundary Detection of Very High Resolution Images Using Mask R-CNN

Anaraki, Neda Rahimpour, Azadbakht, Alireza, Tahmasbi, Maryam, Farahani, Hadi, Kheradpisheh, Saeed Reza, Javaheri, Alireza

arXiv.org Artificial Intelligence

Recently, there has been a high demand for accelerating and improving the detection of automatic cadastral mapping. As this problem is in its starting point, there are many methods of computer vision and deep learning that have not been considered yet. In this paper, we focus on deep learning and provide three geometric post-processing methods that improve the quality of the work. Our framework includes two parts, each of which consists of a few phases. Our solution to this problem uses instance segmentation. In the first part, we use Mask R-CNN with the backbone of pre-trained ResNet-50 on the ImageNet dataset. In the second phase, we apply three geometric post-processing methods to the output of the first part to get better overall output. Here, we also use computational geometry to introduce a new method for simplifying lines which we call it pocket-based simplification algorithm. For evaluating the quality of our solution, we use popular formulas in this field which are recall, precision and F-score. The highest recall we gain is 95 percent which also maintains high Precision of 72 percent. This resulted in an F-score of 82 percent. Implementing instance segmentation using Mask R-CNN with some geometric post-processes to its output gives us promising results for this field. Also, results show that pocket-based simplification algorithms work better for simplifying lines than Douglas-Puecker algorithm.


The end of the stock images industry.

#artificialintelligence

Stock photos are the bane of many marketers and designers. It is a huge part of the life of content marketers, graphic designers and copywriters. They command high fees for usage and are able to justify them due to the fact that the majority of the users of stock images are enterprises. One of the first examples of a stock photo was circa 1920 when American photographer H. Armstrong Roberts ensured that the people photographed in Group in Front of Tri-Motor Airplane all signed model releases. This allowed the photograph and others like it to be commercially viable.


AI Upscaling And The Future Of Content Delivery - AI Summary

#artificialintelligence

The technology, which has already been employed by several notable PC games over the last few years, uses machine learning to upscale rendered images in real-time. So rather than tasking the GPU with producing a native 4K image, the engine can render the game at a lower resolution and have DLSS make up the difference. In the case of DLSS, NVIDIA trained their neural network by taking low and high resolution images of the same game and having their in-house supercomputer analyze the differences. Combined with motion vector data, the neural network was tasked with not only filling in the necessary visual information to make the low resolution image better approximate the idealistic target, but predict what the next frame of animation might look like. In other words, if you have a computer powerful enough to run a game at 30 FPS in 1920 x 1080, the same computer could potentially reach 60 FPS if the game was rendered at 1280 x 720 and scaled up with DLSS.


PCA Reduced Gaussian Mixture Models with Applications in Superresolution

Hertrich, Johannes, Nguyen, Dang Phoung Lan, Aujol, Jean-Fancois, Bernard, Dominique, Berthoumieu, Yannick, Saadaldin, Abdellativ, Steidl, Gabriele

arXiv.org Machine Learning

Despite the rapid development of computational hardware, the treatment of large and high dimensional data sets is still a challenging problem. This paper provides a twofold contribution to the topic. First, we propose a Gaussian Mixture Model in conjunction with a reduction of the dimensionality of the data in each component of the model by principal component analysis, called PCA-GMM. To learn the (low dimensional) parameters of the mixture model we propose an EM algorithm whose M-step requires the solution of constrained optimization problems. Fortunately, these constrained problems do not depend on the usually large number of samples and can be solved efficiently by an (inertial) proximal alternating linearized minimization algorithm. Second, we apply our PCA-GMM for the superresolution of 2D and 3D material images based on the approach of Sandeep and Jacob. Numerical results confirm the moderate influence of the dimensionality reduction on the overall superresolution result.


Artificial intelligence enhances blurry faces into 'super-resolution images'

The Independent - Tech

Researchers have figured out a way to transform a few dozen pixels into a high resolution image of a face using artificial intelligence. A team from Duke University in the US created an algorithm capable of "imagining" realistic-looking faces from blurry, unrecognisable pictures of people, with eight-times more effectiveness than previous methods. "Never have super-resolution images been created at this resolution before with this much detail," said Duke computer scientist Cynthia Rudin, who led the research. The images generated by the AI do not resemble real people, instead they are faces that look plausibly real. It therefore cannot be used to identify people from low resolution images captured by security cameras.